representational harm
Bias Amplification in Stable Diffusion's Representation of Stigma Through Skin Tones and Their Homogeneity
Wilson, Kyra, Ghosh, Sourojit, Caliskan, Aylin
Text-to-image generators (T2Is) are liable to produce images that perpetuate social stereotypes, especially in regards to race or skin tone. We use a comprehensive set of 93 stigmatized identities to determine that three versions of Stable Diffusion (v1.5, v2.1, and XL) systematically associate stigmatized identities with certain skin tones in generated images. We find that SD XL produces skin tones that are 13.53% darker and 23.76% less red (both of which indicate higher likelihood of societal discrimination) than previous models and perpetuate societal stereotypes associating people of color with stigmatized identities. SD XL also shows approximately 30% less variability in skin tones when compared to previous models and 18.89-56.06% compared to human face datasets. Measuring variability through metrics which directly correspond to human perception suggest a similar pattern, where SD XL shows the least amount of variability in skin tones of people with stigmatized identities and depicts most (60.29%) stigmatized identities as being less diverse than non-stigmatized identities. Finally, SD shows more homogenization of skin tones of racial and ethnic identities compared to other stigmatized or non-stigmatized identities, reinforcing incorrect equivalence of biologically-determined skin tone and socially-constructed racial and ethnic identity. Because SD XL is the largest and most complex model and users prefer its generations compared to other models examined in this study, these findings have implications for the dynamics of bias amplification in T2Is, increasing representational harms and challenges generating diverse images depicting people with stigmatized identities.
- North America > United States > Illinois > Cook County > Chicago (0.04)
- Oceania > Australia (0.04)
- North America > United States > South Carolina > Greenville County > Taylors (0.04)
- (5 more...)
Understanding and Meeting Practitioner Needs When Measuring Representational Harms Caused by LLM-Based Systems
Harvey, Emma, Sheng, Emily, Blodgett, Su Lin, Chouldechova, Alexandra, Garcia-Gathright, Jean, Olteanu, Alexandra, Wallach, Hanna
The NLP research community has made publicly available numerous instruments for measuring representational harms caused by large language model (LLM)-based systems. These instruments have taken the form of datasets, metrics, tools, and more. In this paper, we examine the extent to which such instruments meet the needs of practitioners tasked with evaluating LLM-based systems. Via semi-structured interviews with 12 such practitioners, we find that practitioners are often unable to use publicly available instruments for measuring representational harms. We identify two types of challenges. In some cases, instruments are not useful because they do not meaningfully measure what practitioners seek to measure or are otherwise misaligned with practitioner needs. In other cases, instruments - even useful instruments - are not used by practitioners due to practical and institutional barriers impeding their uptake. Drawing on measurement theory and pragmatic measurement, we provide recommendations for addressing these challenges to better meet practitioner needs.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Europe > Austria > Vienna (0.14)
- Asia > Singapore (0.04)
- (18 more...)
- Research Report (1.00)
- Questionnaire & Opinion Survey (1.00)
- Personal > Interview (1.00)
- Health & Medicine (1.00)
- Information Technology (0.93)
- Law (0.93)
Multi-Group Proportional Representation for Text-to-Image Models
Jung, Sangwon, Oesterling, Alex, Verdun, Claudio Mayrink, Vithana, Sajani, Moon, Taesup, Calmon, Flavio P.
Text-to-image (T2I) generative models can create vivid, realistic images from textual descriptions. As these models proliferate, they expose new concerns about their ability to represent diverse demographic groups, propagate stereotypes, and efface minority populations. Despite growing attention to the "safe" and "responsible" design of artificial intelligence (AI), there is no established methodology to systematically measure and control representational harms in image generation. This paper introduces a novel framework to measure the representation of intersectional groups in images generated by T2I models by applying the Multi-Group Proportional Representation (MPR) metric. MPR evaluates the worst-case deviation of representation statistics across given population groups in images produced by a generative model, allowing for flexible and context-specific measurements based on user requirements. We also develop an algorithm to optimize T2I models for this metric. Through experiments, we demonstrate that MPR can effectively measure representation statistics across multiple intersectional groups and, when used as a training objective, can guide models toward a more balanced generation across demographic groups while maintaining generation quality.
- North America > United States (0.46)
- Asia > South Korea > Seoul > Seoul (0.04)
- Asia > China > Tibet Autonomous Region (0.04)
- Information Technology > Sensing and Signal Processing > Image Processing (1.00)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
A Tale of Two Identities: An Ethical Audit of Human and AI-Crafted Personas
Venkit, Pranav Narayanan, Li, Jiayi, Zhou, Yingfan, Rajtmajer, Sarah, Wilson, Shomir
As LLMs (large language models) are increasingly used to generate synthetic personas--particularly in data-limited domains such as health, privacy, and HCI--it becomes necessary to understand how these narratives represent identity, especially that of minority communities. In this paper, we audit synthetic personas generated by 3 LLMs (GPT4o, Gemini 1.5 Pro, Deepseek v2.5) through the lens of representational harm, focusing specifically on racial identity. Using a mixed-methods approach combining close reading, lexical analysis, and a parameterized creativity framework, we compare 1,512 LLM-generated persona to human-authored responses. Our findings reveal that LLMs disproportionately foreground racial markers, overproduce culturally coded language, and construct personas that are syntactically elaborate yet nar-ratively reductive. These patterns result in a range of so-ciotechnical harms--including stereotyping, exoticism, erasure, and benevolent bias--that are often obfuscated by superficially positive narrations. We formalize this phenomenon as algorithmic othering, where minoritized identities are rendered hypervisible but less authentic. Based on these findings, we offer design recommendations for narrative-aware evaluation metrics and community-centered validation protocols for synthetic identity generation.
- North America > United States > Alaska (0.05)
- North America > Mexico (0.04)
- North America > United States > Pennsylvania > Centre County > University Park (0.04)
- (3 more...)
- Research Report > New Finding (1.00)
- Overview (1.00)
Taxonomizing Representational Harms using Speech Act Theory
Corvi, Emily, Washington, Hannah, Reed, Stefanie, Atalla, Chad, Chouldechova, Alexandra, Dow, P. Alex, Garcia-Gathright, Jean, Pangakis, Nicholas, Sheng, Emily, Vann, Dan, Vogel, Matthew, Wallach, Hanna
Representational harms are widely recognized among fairness-related harms caused by generative language systems. However, their definitions are commonly under-specified. We present a framework, grounded in speech act theory (Austin, 1962), that conceptualizes representational harms caused by generative language systems as the perlocutionary effects (i.e., real-world impacts) of particular types of illocutionary acts (i.e., system behaviors). Building on this argument and drawing on relevant literature from linguistic anthropology and sociolinguistics, we provide new definitions stereotyping, demeaning, and erasure. We then use our framework to develop a granular taxonomy of illocutionary acts that cause representational harms, going beyond the high-level taxonomies presented in previous work. We also discuss the ways that our framework and taxonomy can support the development of valid measurement instruments. Finally, we demonstrate the utility of our framework and taxonomy via a case study that engages with recent conceptual debates about what constitutes a representational harm and how such harms should be measured.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > United States > Massachusetts > Hampshire County > Amherst (0.04)
- (4 more...)
More of the Same: Persistent Representational Harms Under Increased Representation
Mickel, Jennifer, De-Arteaga, Maria, Liu, Leqi, Tian, Kevin
To recognize and mitigate the harms of generative AI systems, it is crucial to consider who is represented in the outputs of generative AI systems and how people are represented. A critical gap emerges when naively improving who is represented, as this does not imply bias mitigation efforts have been applied to address how people are represented. We critically examined this by investigating gender representation in occupation across state-of-the-art large language models. We first show evidence suggesting that over time there have been interventions to models altering the resulting gender distribution, and we find that women are more represented than men when models are prompted to generate biographies or personas. We then demonstrate that representational biases persist in how different genders are represented by examining statistically significant word differences across genders. This results in a proliferation of representational harms, stereotypes, and neoliberalism ideals that, despite existing interventions to increase female representation, reinforce existing systems of oppression.
- North America > United States > Texas > Travis County > Austin (0.14)
- Europe > United Kingdom > Scotland (0.14)
- South America (0.04)
- (6 more...)
- Law (1.00)
- Health & Medicine > Therapeutic Area (0.68)
- Government > Regional Government > North America Government > United States Government (0.68)
- Education > Educational Setting > Higher Education (0.46)
Do Generative AI Models Output Harm while Representing Non-Western Cultures: Evidence from A Community-Centered Approach
Ghosh, Sourojit, Venkit, Pranav Narayanan, Gautam, Sanjana, Wilson, Shomir, Caliskan, Aylin
Our research investigates the impact of Generative Artificial Intelligence (GAI) models, specifically text-to-image generators (T2Is), on the representation of non-Western cultures, with a focus on Indian contexts. Despite the transformative potential of T2Is in content creation, concerns have arisen regarding biases that may lead to misrepresentations and marginalizations. Through a community-centered approach and grounded theory analysis of 5 focus groups from diverse Indian subcultures, we explore how T2I outputs to English prompts depict Indian culture and its subcultures, uncovering novel representational harms such as exoticism and cultural misappropriation. These findings highlight the urgent need for inclusive and culturally sensitive T2I systems. We propose design guidelines informed by a sociotechnical perspective, aiming to address these issues and contribute to the development of more equitable and representative GAI technologies globally. Our work also underscores the necessity of adopting a community-centered approach to comprehend the sociotechnical dynamics of these models, complementing existing work in this space while identifying and addressing the potential negative repercussions and harms that may arise when these models are deployed on a global scale.
- North America > United States > Washington > King County > Seattle (0.14)
- Asia > Pakistan (0.04)
- North America > United States > Pennsylvania > Centre County > University Park (0.04)
- (9 more...)
- Leisure & Entertainment (0.93)
- Health & Medicine (0.68)
The Psychosocial Impacts of Generative AI Harms
Vassel, Faye-Marie, Shieh, Evan, Sugimoto, Cassidy R., Monroe-White, Thema
The rapid emergence of generative Language Models (LMs) has led to growing concern about the impacts that their unexamined adoption may have on the social well-being of diverse user groups. Meanwhile, LMs are increasingly being adopted in K-20 schools and one-on-one student settings with minimal investigation of potential harms associated with their deployment. Motivated in part by real-world/everyday use cases (e.g., an AI writing assistant) this paper explores the potential psychosocial harms of stories generated by five leading LMs in response to open-ended prompting. We extend findings of stereotyping harms analyzing a total of 150K 100-word stories related to student classroom interactions. Examining patterns in LM-generated character demographics and representational harms (i.e., erasure, subordination, and stereotyping) we highlight particularly egregious vignettes, illustrating the ways LM-generated outputs may influence the experiences of users with marginalized and minoritized identities, and emphasizing the need for a critical understanding of the psychosocial impacts of generative AI tools when deployed and utilized in diverse social contexts.
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- Asia > Singapore (0.04)
- Research Report > New Finding (0.68)
- Research Report > Experimental Study (0.46)
- Government (0.93)
- Education > Educational Setting (0.47)
Misgendering and Assuming Gender in Machine Translation when Working with Low-Resource Languages
Ghosh, Sourojit, Chatterjee, Srishti
This chapter focuses on gender-related errors in machine translation (MT) in the context of low-resource languages. We begin by explaining what low-resource languages are, examining the inseparable social and computational factors that create such linguistic hierarchies. We demonstrate through a case study of our mother tongue Bengali, a global language spoken by almost 300 million people but still classified as low-resource, how gender is assumed and inferred in translations to and from the high(est)-resource English when no such information is provided in source texts. We discuss the postcolonial and societal impacts of such errors leading to linguistic erasure and representational harms, and conclude by discussing potential solutions towards uplifting languages by providing them more agency in MT conversations.
- Asia > China (0.15)
- North America > United States > New York > New York County > New York City (0.05)
- Asia > Pakistan (0.05)
- (23 more...)
- Government > Regional Government (0.68)
- Law (0.68)
Interview with Aylin Caliskan: AI ethics
In 2023, Aylin Caliskan was recognized as one of the 100 Brilliant Women in AI Ethics. At this year's International Joint Conference on Artificial Intelligence (IJCAI 2023) she gave an IJCAI Early Career Spotlight talk about her work. I met with Aylin at the conference and chatted to her about AI ethics. We spoke about bias in generative AI tools and the associated research and societal challenges. Andrea Rafai: We've seen generative AI tools become mainstream recently.